4 research outputs found
Beyond Generating Code: Evaluating GPT on a Data Visualization Course
This paper presents an empirical evaluation of the performance of the
Generative Pre-trained Transformer (GPT) model in Harvard's CS171 data
visualization course. While previous studies have focused on GPT's ability to
generate code for visualizations, this study goes beyond code generation to
evaluate GPT's abilities in various visualization tasks, such as data
interpretation, visualization design, visual data exploration, and insight
communication. The evaluation utilized GPT-3.5 and GPT-4 to complete
assignments of CS171, and included a quantitative assessment based on the
established course rubrics, a qualitative analysis informed by the feedback of
three experienced graders, and an exploratory study of GPT's capabilities in
completing border visualization tasks. Findings show that GPT-4 scored 80% on
quizzes and homework, and TFs could distinguish between GPT- and
human-generated homework with 70% accuracy. The study also demonstrates GPT's
potential in completing various visualization tasks, such as data cleanup,
interaction with visualizations, and insight communication. The paper concludes
by discussing the strengths and limitations of GPT in data visualization,
potential avenues for incorporating GPT in broader visualization tasks, and the
need to redesign visualization education.Comment: vis short papg
Dense 4D nanoscale reconstruction of living brain tissue
Three-dimensional (3D) reconstruction of living brain tissue down to an individual synapse level would create opportunities for decoding the dynamics and structure–function relationships of the brain’s complex and dense information processing network; however, this has been hindered by insufficient 3D resolution, inadequate signal-to-noise ratio and prohibitive light burden in optical imaging, whereas electron microscopy is inherently static. Here we solved these challenges by developing an integrated optical/machine-learning technology, LIONESS (live information-optimized nanoscopy enabling saturated segmentation). This leverages optical modifications to stimulated emission depletion microscopy in comprehensively, extracellularly labeled tissue and previous information on sample structure via machine learning to simultaneously achieve isotropic super-resolution, high signal-to-noise ratio and compatibility with living tissue. This allows dense deep-learning-based instance segmentation and 3D reconstruction at a synapse level, incorporating molecular, activity and morphodynamic information. LIONESS opens up avenues for studying the dynamic functional (nano-)architecture of living brain tissue
Saturated reconstruction of living brain tissue
Complex wiring between neurons underlies the information-processing network enabling all brain functions, including cognition and memory. For understanding how the network is structured, processes information, and changes over time, comprehensive visualization of the architecture of living brain tissue with its cellular and molecular components would open up major opportunities. However, electron microscopy (EM) provides nanometre-scale resolution required for full in-silico reconstruction1–5, yet is limited to fixed specimens and static representations. Light microscopy allows live observation, with super-resolution approaches6–12 facilitating nanoscale visualization, but comprehensive 3D-reconstruction of living brain tissue has been hindered by tissue photo-burden, photobleaching, insufficient 3D-resolution, and inadequate signal-to-noise ratio (SNR). Here we demonstrate saturated reconstruction of living brain tissue. We developed an integrated imaging and analysis technology, adapting stimulated emission depletion (STED) microscopy6,13 in extracellularly labelled tissue14 for high SNR and near-isotropic resolution. Centrally, a two-stage deep-learning approach leveraged previously obtained information on sample structure to drastically reduce photo-burden and enable automated volumetric reconstruction down to single synapse level. Live reconstruction provides unbiased analysis of tissue architecture across time in relation to functional activity and targeted activation, and contextual understanding of molecular labelling. This adoptable technology will facilitate novel insights into the dynamic functional architecture of living brain tissue